Tuning Variable Selection Procedures by Adding Noise
نویسندگان
چکیده
Many variable selection methods for linear regression depend critically on tuning parameters that control the performance of the method, e.g., “entry” and“stay” significance levels in forward and backward selection. However, most methods do not adapt the tuning parameters to particular data sets. We propose a general strategy for adapting variable selection tuning parameters that effectively estimates the tuning parameters so that the selection method avoids overfitting and underfitting. The strategy is based on the principle that underfitting and overfitting can be directly observed in estimates of the error variance after 1) adding controlled amounts of additional independent noise to the response variable and 2) then running a variable selection method. It is related to the simulation technique SIMEX found in the measurement error literature. We focus on forward selection because of its simplicity and ability to handle large numbers of explanatory variables. Monte Carlo studies show that the new method compares favorably with established methods. Xiaohui Luo is Biometrician, Clinical Biostatistics, Merck Research Laboratories, Rahway, NJ 07065-0900 (Email: edmund [email protected]). Leonard A. Stefanski and Dennis D. Boos are Professors, Department of Statistics, North Carolina State University, Raleigh, NC 27695-8203 (E-mail: [email protected], [email protected]). 1
منابع مشابه
Comparison of aperiodic stochastic resonance in a bistable system realized by adding noise and by tuning system parameters.
Two methods of realizing aperiodic stochastic resonance (ASR) by adding noise and tuning system parameters in a bistable system, after a scale transformation, can be compared in a real parameter space. In this space, the resonance point of ASR via adding noise denotes the extremum of a line segment, whereas the method of tuning system parameters presents the extrema of a parameter plane. We dem...
متن کاملHigh SNR consistent compressive sensing
High signal to noise ratio (SNR) consistency of model selection criteria in linear regression models has attracted a lot of attention recently. However, most of the existing literature on high SNR consistency deals with model order selection. Further, the limited literature available on the high SNR consistency of subset selection procedures (SSPs) is applicable to linear regression with full r...
متن کاملA Note on the Lasso and Related Procedures in Model Selection
The Lasso, the Forward Stagewise regression and the Lars are closely related procedures recently proposed for linear regression problems. Each of them can produce sparse models and can be used both for estimation and variable selection. In practical implementations these algorithms are typically tuned to achieve optimal prediction accuracy. We show that, when the prediction accuracy is used as ...
متن کاملBayesian Variable Selections for Probit Models with Componentwise Gibbs Samplers
For variable selection to binary response regression, stochastic search variable selection and Bayesian Lasso have recently been popular. However, these two variable selection methods suffer from heavy computation burden caused by hyperparameter tuning and by matrix inversions, especially when the number of covariates is large. Therefore, this article incorporates the componenetwise Gibbs sampl...
متن کاملDeterministic and Random Selection of Variables in Local Search for SAT
The adaptive noise mechanism [2] was introduced in Novelty+ [1] to automatically adapt noise settings during the search, yielding the algorithm adaptNovelty+. The local search algorithm GWSAT deterministically selects the best promising decreasing variable to flip, if such variables exist [3]. However, the performance of GWSAT still depends on static noise settings. To avoid manual noise tuning...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- Technometrics
دوره 48 شماره
صفحات -
تاریخ انتشار 2006